Statistical Regularization and Learning Theory Lecture : 12 Complexity Regularization in Regression

نویسنده

  • David Winters
چکیده

Example 1 To illustrate the distinction between classification and regression, consider a simple, scalar signal plus noise problem. Consider Yi = θ +Wi, i = 1, . . . , n, where θ is a fixed unknown scalar parameter and the Wi are independent, zero-mean, unit variance random variables. Let Ȳ = 1/n ∑n i=1 Yi. Then, according to the Central Limit Theorem, Ȳ is distributed approximately N(θ, 1/n). A simple tail-bound on the Gaussian distribution gives us

برای دانلود رایگان متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

ثبت نام

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

منابع مشابه

On Different Facets of Regularization Theory

This review provides a comprehensive understanding of regularization theory from different perspectives, emphasizing smoothness and simplicity principles. Using the tools of operator theory and Fourier analysis, it is shown that the solution of the classical Tikhonov regularization problem can be derived from the regularized functional defined by a linear differential (integral) operator in the...

متن کامل

Learning Rates of lq Coefficient Regularization Learning with Gaussian Kernel

Regularization is a well-recognized powerful strategy to improve the performance of a learning machine and l(q) regularization schemes with 0 < q < ∞ are central in use. It is known that different q leads to different properties of the deduced estimators, say, l(2) regularization leads to a smooth estimator, while l(1) regularization leads to a sparse estimator. Then how the generalization capa...

متن کامل

Elastic-net regularization in learning theory

Within the framework of statistical learning theory we analyze in detail the so-called elastic-net regularization scheme proposed by Zou and Hastie [H. Zou, T. Hastie, Regularization and variable selection via the elastic net, J. R. Stat. Soc. Ser. B, 67(2) (2005) 301–320] for the selection of groups of correlated variables. To investigate the statistical properties of this scheme and in partic...

متن کامل

Statistical Likelihood Representations of Prior Knowledge in Machine Learning

We show that maximum a posteriori (MAP) statistical methods can be used in nonparametric machine learning problems in the same way as their current applications in parametric statistical problems, and give some examples of applications. This MAPN (MAP for nonparametric machine learning) paradigm can also reproduce much more transparently the same results as regularization methods in machine lea...

متن کامل

Optimal Computational and Statistical Rates of Convergence for Sparse Nonconvex Learning Problems.

We provide theoretical analysis of the statistical and computational properties of penalized M-estimators that can be formulated as the solution to a possibly nonconvex optimization problem. Many important estimators fall in this category, including least squares regression with nonconvex regularization, generalized linear models with nonconvex regularization and sparse elliptical random design...

متن کامل

ذخیره در منابع من


  با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید

عنوان ژورنال:

دوره   شماره 

صفحات  -

تاریخ انتشار 2004